Posted: Sept. 23, 2025
What happens when you sit today’s most advanced AI models down with a stack of bar exam questions? That’s exactly what we wanted to find out. Our team tested 59 different large language models (LLMs)—from companies like OpenAI, Google, Anthropic, Meta, and more on a practice Multistate Bar Exam (MBE) questions. For the non-lawyers, the MBE section of the bar exam is 200 multiple choice questions that equally cover Civil Procedure, Contracts, Evidence, Torts, Constitutional Law, Criminal Law and Procedure, and Real Property. The multiple choice format makes it much easier to quickly evaluate whether an A.I. is answering correctly. You can explore the data at the data dashboard located at https://ai-mbe-study.streamlit.app/.
The results? Pretty eye-opening.
Some legal AI startups boast sky-high valuations, but they may have a hard time proving their edge when free or cheap base models are already close to their level of accuracy. The market may soon ask: what are we really paying for?
AI has come a long way since ChatGPT’s launch in 2022, and it’s only getting better. But like any tool in law, it needs careful use, ongoing scrutiny, and a healthy dose of human judgment.
If you’re curious to dive deeper, you can check out the full study and methodology here. (The paper was finalized in May 2025 so only includes the models up to that point.)
The code and data are all open source. Check out all of the data at https://ai-mbe-study.streamlit.app/.